Low bit-rate feature vector compression using transform coding and non-uniform bit allocation

نویسندگان

  • Ben P. Milner
  • Xu Shao
چکیده

This paper presents a novel method for the low bit-rate compression of a feature vector stream with particular application to distributed speech recognition. The scheme operates by grouping feature vectors into non-overlapping blocks and applying a transformation to give a more compact matrix representation. Both Karhunen-Loeve and discrete cosine transforms are considered. Following transformation, higher-order columns of the matrix can be removed without loss in recognition performance. The number of bits allocated to the remaining elements in the matrix is determined automatically using a measure of their relative information content. Analysis of the amplitude distribution of the elements indicates that non-linear quantisation is more appropriate than linear quantisation. Comparative results, based on both spectral distortion and speech recognition accuracy, confirm this. Speech recognition tests using the ETSI Aurora database demonstrate that compression to bits rates of 2400bps, 1200bps and 800bps has very little effect on recognition accuracy. For example at a bit rate of 1200bps, recognition accuracy is 98.0% compared to 98.6% with no compression.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Non-linear Compression of Fea Transform Coding and Non-unif

This paper uses transform coding for compressing feature vectors in distributed speech recognition applications. Feature vectors are first grouped together into non-overlapping blocks and a transformation applied. A non-uniform allocation of bits to the elements of the resultant matrix is based on their relative information content. Analysis of the amplitude distribution of these elements indic...

متن کامل

Improved Modeling and Quantization Methods for Speech Coding

With the advent of 3G Wireless standards and subsequent bandwidth expansion, there is a clear need to design high quality, low complexity compression schemes which are bit-efficient. We have proposed a computationally efficient, high quality, vector quantization scheme based on a parametric probability density function (PDF). In this scheme, speech line spectral frequencies (LSF) are modeled as...

متن کامل

Weighted universal transform coding: universal image compression with the Karhunen-Loeve transform

We introduce a two-stage universal transform code for image compression. The code combines Karhunen-Lo eve transform coding with weighted universal bit allocation (WUBA) 1] in a two-stage algorithm analogous to the algorithm for weighted universal vector quantization (WUVQ) 2, 3]. The encoder uses a collection of transform / bit allocation pairs rather than a single transform / bit allocation p...

متن کامل

A hybrid vector quantizer for enhanced image pyramid coding with application to volumetric image compression in confocal microscopy [5637-4]

Three-dimensional image compression methods outperform their two-dimensional counterparts in the sense of higher rate-distortion performance for compressing volumetric image data. The state-of-the-art transform-based 3D compressors, such as 3D-SPIHT and 3D-DCT, are characterized for their rate control ability, where the qualities of the image, although are adjustable with respect to rates, are ...

متن کامل

A hybrid vector quantizer for enhanced image pyramid coding with application to volumetric image compression in confocal microscopy

Three-dimensional image compression methods outperform their two-dimensional counterparts in the sense of higher rate-distortion performance for compressing volumetric image data. The state-of-the-art transform-based 3D compressors, such as 3D-SPIHT and 3D-DCT, are characterized for their rate control ability, where the qualities of the image, although are adjustable with respect to rates, are ...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2003